Goto

Collaborating Authors

 episodic restless bandit problem


Regret Bounds for Thompson Sampling in Episodic Restless Bandit Problems

Neural Information Processing Systems

Restless bandit problems are instances of non-stationary multi-armed bandits. These problems have been studied well from the optimization perspective, where the goal is to efficiently find a near-optimal policy when system parameters are known. However, very few papers adopt a learning perspective, where the parameters are unknown. In this paper, we analyze the performance of Thompson sampling in episodic restless bandits with unknown parameters. We consider a general policy map to define our competitor and prove an $\tilde{\bigO}(\sqrt{T})$ Bayesian regret bound. Our competitor is flexible enough to represent various benchmarks including the best fixed action policy, the optimal policy, the Whittle index policy, or the myopic policy.


Reviews: Regret Bounds for Thompson Sampling in Episodic Restless Bandit Problems

Neural Information Processing Systems

The restrictive assumption of restarting (which also significantly simplifies the regret analysis) was not mentioned. Note that the work by Liu, et.


Reviews: Regret Bounds for Thompson Sampling in Episodic Restless Bandit Problems

Neural Information Processing Systems

The reviewers liked this paper, and I did as well. One thought is whether or not Exp4 can be adapted to this setting. The translation is not immediate by any means, but perhaps this is worth thinking about. Please take the reviewers suggestions into consideration for the final version as promised in your response.


Regret Bounds for Thompson Sampling in Episodic Restless Bandit Problems

Neural Information Processing Systems

Restless bandit problems are instances of non-stationary multi-armed bandits. These problems have been studied well from the optimization perspective, where the goal is to efficiently find a near-optimal policy when system parameters are known. However, very few papers adopt a learning perspective, where the parameters are unknown. In this paper, we analyze the performance of Thompson sampling in episodic restless bandits with unknown parameters. We consider a general policy map to define our competitor and prove an \tilde{\bigO}(\sqrt{T}) Bayesian regret bound.


Regret Bounds for Thompson Sampling in Episodic Restless Bandit Problems

Jung, Young Hun, Tewari, Ambuj

Neural Information Processing Systems

Restless bandit problems are instances of non-stationary multi-armed bandits. These problems have been studied well from the optimization perspective, where the goal is to efficiently find a near-optimal policy when system parameters are known. However, very few papers adopt a learning perspective, where the parameters are unknown. In this paper, we analyze the performance of Thompson sampling in episodic restless bandits with unknown parameters. We consider a general policy map to define our competitor and prove an $\tilde{\bigO}(\sqrt{T})$ Bayesian regret bound. Our competitor is flexible enough to represent various benchmarks including the best fixed action policy, the optimal policy, the Whittle index policy, or the myopic policy.